Multi-reward reinforcement learning based development of inter-atomic potential models for silica

نویسندگان

چکیده

Abstract Silica is an abundant and technologically attractive material. Due to the structural complexities of silica polymorphs coupled with subtle differences in Si–O bonding characteristics, development accurate models predict structure, energetics properties remain challenging. Current for range from computationally efficient Buckingham formalisms (BKS, CHIK, Soules) reactive (ReaxFF) more recent machine-learned potentials that are flexible but costly. Here, we introduce improved formalism parameterization BKS model via a multireward reinforcement learning (RL) using experimental training dataset. Our concurrently captures energetics, density, equation state, elastic constants quartz (equilibrium) as well 20 other metastable polymorphs. We also assess its ability capturing amorphous highlight limitations BKS-type functional forms simultaneously crystal properties. demonstrate ways improve flexibility formalism, ML-BKS, outperforms existing empirical on-par recently developed 50 100 times expensive Gaussian approximation potential (GAP) structure silica.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Potential Based Reward Shaping for Hierarchical Reinforcement Learning

Hierarchical Reinforcement Learning (HRL) outperforms many ‘flat’ Reinforcement Learning (RL) algorithms in some application domains. However, HRL may need longer time to obtain the optimal policy because of its large action space. Potential Based Reward Shaping (PBRS) has been widely used to incorporate heuristics into flat RL algorithms so as to reduce their exploration. In this paper, we inv...

متن کامل

Potential-based reward shaping for knowledge-based, multi-agent reinforcement learning

Reinforcement learning is a robust artificial intelligence solution for agents required to act in an environment, making their own decisions on how to behave. Typically an agent is deployed alone with no prior knowledge, but if given sufficient time, a suitable state representation and an informative reward function is guaranteed to learn how to maximise its long term reward. Incorporating doma...

متن کامل

Plan-based reward shaping for multi-agent reinforcement learning

Recent theoretical results have justified the use of potential-based reward shaping as a way to improve the performance of multi-agent reinforcement learning (MARL). However, the question remains of how to generate a useful potential function. Previous research demonstrated the use of STRIPS operator knowledge to automatically generate a potential function for single-agent reinforcement learnin...

متن کامل

Abstract MDP Reward Shaping for Multi-Agent Reinforcement Learning

MDP Reward Shaping for Multi-Agent Reinforcement Learning Kyriakos Efthymiadis, Sam Devlin and Daniel Kudenko Department of Computer Science, The University of York, UK Abstract. Reward shaping has been shown to significantly improve an agent’s performance in reinforcement learning. As attention is shifting from tabula-rasa approaches to methods where some heuristic domain knowledge can be give...

متن کامل

Multi-Time Models for Reinforcement Learning

Reinforcement learning can be used not only to predict rewards, but also to predict states, i.e. to learn a model of the world's dynamics. Models can be deened at diierent levels of temporal abstraction. Multi-time models are models that focus on predicting what will happen, rather than when a certain event will take place. Based on multi-time models, we can deene abstract actions, which enable...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: npj computational materials

سال: 2023

ISSN: ['2057-3960']

DOI: https://doi.org/10.1038/s41524-023-01074-9